Appearance-Based Gaze Estimation Method Using Static Transformer Temporal Differential Network
نویسندگان
چکیده
Gaze behavior is important and non-invasive human–computer interaction information that plays an role in many fields—including skills transfer, psychology, interaction. Recently, improving the performance of appearance-based gaze estimation, using deep learning techniques, has attracted increasing attention: however, several key problems these deep-learning-based estimation methods remain. Firstly, feature fusion stage not fully considered: existing simply concatenate different obtained features into one feature, without considering their internal relationship. Secondly, dynamic can be difficult to learn, because unstable extraction process ambiguously defined features. In this study, we propose a novel method consider problems. We static transformer module (STM), which uses multi-head self-attention mechanism fuse fine-grained eye coarse-grained facial Additionally, innovative recurrent neural network (RNN) cell—that is, temporal differential (TDM)—which used extract integrated STM TDM with (STTDN). evaluated STTDN performance, two publicly available datasets (MPIIFaceGaze Eyediap), demonstrated effectiveness TDM. Our results show proposed outperformed state-of-the-art methods, including Eyediap (by 2.9%).
منابع مشابه
Gaze estimation using appearance models
A simple gaze estimation procedure based on appearance models is described. The method uses minimum norm linear regression to directly relate eye appearance to gaze direction. No explicit feature extraction is required. The method is fast, requiring only linear operations on the data. It is also accurate, in spite of using only a small amount of training data.
متن کاملAppearance-based Eye Gaze Estimation
We present a method for estimating eye gaze direction, which represents a departure from conventional eye gaze estimation methods, the majority of which are based on tracking specific optical phenomena like corneal reflection and the Purkinje images. We employ an appearance manifold model, but instead of using a densely sampled spline to perform the nearest manifold point query, we retain the o...
متن کاملMPIIGaze: Real-World Dataset and Deep Appearance-Based Gaze Estimation
Learning-based methods are believed to work well for unconstrained gaze estimation, i.e. gaze estimation from a monocular RGB camera without assumptions regarding user, environment, or camera. However, current gaze datasets were collected under laboratory conditions and methods were not evaluated across multiple datasets. Our work makes three contributions towards addressing these limitations. ...
متن کاملA Head Pose-free Approach for Appearance-based Gaze Estimation
(a) (b) (c) Figure 1: 2-D illustration of relationship between gaze direction and head pose. (a) Under a fixed head pose (r0, t0), gaze direction α can be estimated from appearance. (b) To obtain α under another head pose (r̂, t̂), the estimated α ′ should be corrected because of captured eye appearance distortion. (c) Under head pose (r̂, t̂), gaze direction under WCS should be further compensated...
متن کاملActive Appearance Models for Gaze Estimation
This thesis describes the implementation of and the experimentation with a gaze estimation system based on an approach called the Active Appearance Model (AAM). Active Appearance Models are generative models of a certain visual phenomenon. They can be used to model specific patterns of variability in shape and grey-level appearance, which in turn can be used directly in image interpretation. Ou...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Mathematics
سال: 2023
ISSN: ['2227-7390']
DOI: https://doi.org/10.3390/math11030686